Hybrid fuzzy AHP–TOPSIS approach to prioritizing solutions for inverse reinforcement learning

نویسندگان

چکیده

Abstract Reinforcement learning (RL) techniques nurture building up solutions for sequential decision-making problems under uncertainty and ambiguity. RL has agents with a reward function that interacts dynamic environment to find out an optimal policy. There are associated like the should be specified in advance, design difficulties unable handle large complex problems, etc. This led development of inverse reinforcement (IRL). IRL also suffers from many real life robust functions, ill-posed etc., different have been proposed solve these maximum entropy, support multiple rewards non-linear majorly eight problems. paper hybrid fuzzy AHP–TOPSIS approach prioritize while implementing IRL. Fuzzy Analytical Hierarchical Process (FAHP) is used get weights identified The relative accuracy root-mean-squared error using FAHP 97.74 0.0349, respectively. Technique Order Preference by Similarity Ideal Solution (TOPSIS) uses solutions. most significant problem implementation ‘lack functions’ weighting 0.180, whereas solution ‘Supports policy functions along stochastic transition models’ having closeness coefficient (CofC) value 0.967156846.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Cascaded Supervised Learning Approach to Inverse Reinforcement Learning

This paper considers the Inverse Reinforcement Learning (IRL) problem, that is inferring a reward function for which a demonstrated expert policy is optimal. We propose to break the IRL problem down into two generic Supervised Learning steps: this is the Cascaded Supervised IRL (CSI) approach. A classification step that defines a score function is followed by a regression step providing a rewar...

متن کامل

Inverse Reinforcement Learning for Marketing

Learning customer preferences from an observed behaviour is an important topic in the marketing literature. Structural models typically model forward-looking customers or firms as utility-maximizing agents whose utility is estimated using methods of Stochastic Optimal Control. We suggest an alternative approach to study dynamic consumer demand, based on Inverse Reinforcement Learning (IRL). We ...

متن کامل

Recurrent Reinforcement Learning: A Hybrid Approach

Successful applications of reinforcement learning in real-world problems often require dealing with partially observable states. It is in general very challenging to construct and infer hidden states as they often depend on the agent’s entire interaction history and may require substantial domain knowledge. In this work, we investigate a deep-learning approach to learning the representation of ...

متن کامل

Hierarchical Reinforcement Learning: A Hybrid Approach

In this thesis we investigate the relationships between the symbolic and subsymbolic methods used for controlling agents by artificial intelligence, focusing in particular on methods that learn. In light of the strengths and weaknesses of each approach, we propose a hybridisation of symbolic and subsymbolic methods to capitalise on the best features of each. We implement such a hybrid system, c...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Complex & Intelligent Systems

سال: 2022

ISSN: ['2198-6053', '2199-4536']

DOI: https://doi.org/10.1007/s40747-022-00807-5